Hebbian Learning of Bayes Optimal Decisions
نویسندگان
چکیده
Uncertainty is omnipresent when we perceive or interact with our environment, and the Bayesian framework provides computational methods for dealing with it. Mathematical models for Bayesian decision making typically require datastructures that are hard to implement in neural networks. This article shows that even the simplest and experimentally best supported type of synaptic plasticity, Hebbian learning, in combination with a sparse, redundant neural code, can in principle learn to infer optimal Bayesian decisions. We present a concrete Hebbian learning rule operating on log-probability ratios. Modulated by reward-signals, this Hebbian plasticity rule also provides a new perspective for understanding how Bayesian inference could support fast reinforcement learning in the brain. In particular we show that recent experimental results by Yang and Shadlen [1] on reinforcement learning of probabilistic inference in primates can be modeled in this way.
منابع مشابه
Hebbian learning for deciding optimally among many alternatives (almost)
Reward-maximizing performance and neurally plausible mechanisms for achieving it have been completely characterized for a general class of two-alternative decision making tasks, and data suggest that humans can implement the optimal procedure. A greater number of alternatives complicates the analysis, but here too, analytical approximations to optimality that are physically and psychologically ...
متن کاملHebbian Plasticity for Improving Perceptual Decisions
Shibata et al. reported that humans could learn to repeatedly evoke a stimulus-associated functional magnetic resonance imaging (fMRI) activity pattern in visual areas V1/V2 through which visual perceptual learning was achieved without stimulus presentation. Contrary to their attribution of visual improvements to neuroplasticity in adult V1/V2, our Hebbian learning interpretation of these data ...
متن کاملOptimal Structural Nested Models for Optimal Sequential Decisions
I describe two new methods for estimating the optimal treatment regime (equivalently, protocol, plan or strategy) from very high dimesional observational and experimental data: (i) g-estimation of an optimal double-regime structural nested mean model (drSNMM) and (ii) g-estimation of a standard single regime SNMM combined with sequential dynamicprogramming (DP) regression. These methods are com...
متن کاملUsing Machine Learning for Operational Decisions in Adversarial Environments
Classical supervised learning assumes that training data is representative of the data expected to be observed in the future. This assumption is clearly violated when an intelligent adversary actively tries to deceive the learner by generating instances very different from those previously seen. The literature on adversarial machine learning aims to address this problem, but often assumes const...
متن کاملDo Hebbian synapses estimate entropy?
Hebbian learning is one of the mainstays of biologically inspired neural processing. Hebb’s rule is biologically plausible, and it has been extensively utilized in both computational neuroscience and in unsupervised training of neural systems. In these fields, Hebbian learning became synonymous for correlation learning. But it is known that correlation is a second order statistic of the data, s...
متن کامل